Information-theoretics Based Error-metrics for Gradient Descent Learning in Neutral Networks
نویسندگان
چکیده
Conventionally, square error (SE) and/or relative ent ropy (RE) error functions defined over a training set are adop ted towa rds optimization of gradient descent learnings in neur al networks. As an alternative, a set of divergence (or dist ance) measur es can be specified in the inform ationtheoretic plane that funct iona lly have pragmatic values similar to (or improved upon) The SE or RE metrics. KullbackLeibler (KL) , Jensen (J ), and Jensen-Sh ann on (JS) meas ures are suggested as possible inform ation-theoretic error-met ric candidates that are defined and derived explicit ly. Both convent ional SE / RE measures , as well as the prop osed inform ation-theoreti c error-met rics, are applied to t rain a multilayer perceptron topology. This is done in order to elucida te t heir relative efficacy in deciding the perform ance of the network as evidenced from the convergence rat es and tr aining t imes involved. Pertinent simulation results are present ed and discussed.
منابع مشابه
Handwritten Character Recognition using Modified Gradient Descent Technique of Neural Networks and Representation of Conjugate Descent for Training Patterns
The purpose of this study is to analyze the performance of Back propagation algorithm with changing training patterns and the second momentum term in feed forward neural networks. This analysis is conducted on 250 different words of three small letters from the English alphabet. These words are presented to two vertical segmentation programs which are designed in MATLAB and based on portions (1...
متن کاملGeoid Determination Based on Log Sigmoid Function of Artificial Neural Networks: (A case Study: Iran)
A Back Propagation Artificial Neural Network (BPANN) is a well-known learning algorithmpredicated on a gradient descent method that minimizes the square error involving the networkoutput and the goal of output values. In this study, 261 GPS/Leveling and 8869 gravity intensityvalues of Iran were selected, then the geoid with three methods “ellipsoidal stokes integral”,“BPANN”, and “collocation” ...
متن کاملDesigning stable neural identifier based on Lyapunov method
The stability of learning rate in neural network identifiers and controllers is one of the challenging issues which attracts great interest from researchers of neural networks. This paper suggests adaptive gradient descent algorithm with stable learning laws for modified dynamic neural network (MDNN) and studies the stability of this algorithm. Also, stable learning algorithm for parameters of ...
متن کاملNovel Radial Basis Function Neural Networks based on Probabilistic Evolutionary and Gaussian Mixture Model for Satellites Optimum Selection
In this study, two novel learning algorithms have been applied on Radial Basis Function Neural Network (RBFNN) to approximate the functions with high non-linear order. The Probabilistic Evolutionary (PE) and Gaussian Mixture Model (GMM) techniques are proposed to significantly minimize the error functions. The main idea is concerning the various strategies to optimize the procedure of Gradient ...
متن کاملDirect adaptive output tracking control using multilayered neural networks - Control Theory and Applications [see also IEE Proceedings-Control Theory and Applications], IEE Pr
Multilayered neural networks are used to construct nonlinear learning control systems for a class of unknown nonlinear systems in a canonical form. An adaptive output tracking architecture is proposed using the outputs of the two three-layered neutral networks which are trained to approximate the unknown nonlinear plant to any desired degree of accuracy by using the modified back-propagation te...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Complex Systems
دوره 9 شماره
صفحات -
تاریخ انتشار 1995